# 8-bit quantized inference
Qwen3 30B A3B Gptq 8bit
Apache-2.0
Qwen3 30B A3B is a large language model that has undergone 8-bit quantization using the GPTQ method, suitable for efficient inference scenarios.
Large Language Model
Transformers

Q
btbtyler09
301
2
Josiefied Qwen3 8B Abliterated V1 8bit
An optimized 8-bit quantized version of Qwen3-8B, designed for efficient inference on the MLX framework
Large Language Model
J
mlx-community
450
1
Stable Diffusion 3.5 Large Turbo Openvino 8bit
Other
An 8-bit quantized version of Stable Diffusion 3.5 Large Turbo, converted to OpenVINO format for efficient text-to-image generation tasks.
Text-to-Image English
S
AIFunOver
28
1
LLAMA 3 Quantized
MIT
The 8-bit quantized version of Meta Llama 3 - 8B Instruct large language model, reducing model size and improving inference speed, suitable for deployment on resource-constrained devices.
Large Language Model
Transformers

L
Kameshr
18
9
Pinokio V1.0
Apache-2.0
Pinokio_v1.0 is a text generation model supporting English and Turkish, released under the Apache-2.0 license.
Large Language Model
Transformers Supports Multiple Languages

P
Eurdem
2,810
4
Featured Recommended AI Models